Automatic Taxonomy Extraction in Different Languages Using Wikipedia and Minimal Language-Specific Information
نویسندگان
چکیده
منابع مشابه
Language - and domain - independent text mining
Aalto University, P.O. Box 11000, FI-00076 Aalto www.aalto.fi Author Mari-Sanna Paukkeri Name of the doctoral dissertation Languageand domain-independent text mining Publisher School of Science Unit Department of Information and Computer Science Series Aalto University publication series DOCTORAL DISSERTATIONS 137/2012 Field of research Computer and Information Science Manuscript submitted 4 Ma...
متن کاملBootstrapping Term Extractors for Multiple Languages
Terminology extraction resources are needed for a wide range of human language technology applications, including knowledge management, information extraction, semantic search, cross-language information retrieval and automatic and assisted translation. We report a low cost method for creating terminology extraction resources for 21 non-English EU languages. Using parallel corpora and a project...
متن کاملPazienza University of Roma Tor Vergata , Italy Armando Stellato University of Roma Tor Vergata , Italy Semi - Automatic Ontology Development : Processes and Resources
The collection of the specialized vocabulary of a particular domain (terminology) is an important initial step of creating formalized domain knowledge representations (ontologies). Terminology Extraction (TE) aims at automating this process by collecting the relevant domain vocabulary from existing lexical resources or collections of domain texts. In this chapter, the authors address the extrac...
متن کاملInformation Extraction from Wikipedia Using Pattern Learning
In this paper we present solutions for the crucial task of extracting structured information from massive free-text resources, such as Wikipedia, for the sake of semantic databases serving upcoming Semantic Web technologies. We demonstrate both a verb frame-based approach using deep natural language processing techniques with extraction patterns developed by human knowledge experts and machine ...
متن کاملAutomatic Population of Structured Knowledge Bases via Natural Language Processing
The Web has evolved into a huge mine of knowledge carved in different forms, the predominant one still being the free-text document. This motivates the need for Intelligent Web-reading Agents: hypothetically, they would skim through disparate Web sources corpora and generate meaningful structured assertions to fuel Knowledge Bases (KBs). Ultimately, comprehensive KBs, like Wikidata and DBpedia,...
متن کامل